35 research outputs found
Bayesian Active Edge Evaluation on Expensive Graphs
Robots operate in environments with varying implicit structure. For instance,
a helicopter flying over terrain encounters a very different arrangement of
obstacles than a robotic arm manipulating objects on a cluttered table top.
State-of-the-art motion planning systems do not exploit this structure, thereby
expending valuable planning effort searching for implausible solutions. We are
interested in planning algorithms that actively infer the underlying structure
of the valid configuration space during planning in order to find solutions
with minimal effort. Consider the problem of evaluating edges on a graph to
quickly discover collision-free paths. Evaluating edges is expensive, both for
robots with complex geometries like robot arms, and for robots with limited
onboard computation like UAVs. Until now, this challenge has been addressed via
laziness i.e. deferring edge evaluation until absolutely necessary, with the
hope that edges turn out to be valid. However, all edges are not alike in value
- some have a lot of potentially good paths flowing through them, and some
others encode the likelihood of neighbouring edges being valid. This leads to
our key insight - instead of passive laziness, we can actively choose edges
that reduce the uncertainty about the validity of paths. We show that this is
equivalent to the Bayesian active learning paradigm of decision region
determination (DRD). However, the DRD problem is not only combinatorially hard,
but also requires explicit enumeration of all possible worlds. We propose a
novel framework that combines two DRD algorithms, DIRECT and BISECT, to
overcome both issues. We show that our approach outperforms several
state-of-the-art algorithms on a spectrum of planning problems for mobile
robots, manipulators and autonomous helicopters
Adaptive Information Gathering via Imitation Learning
In the adaptive information gathering problem, a policy is required to select
an informative sensing location using the history of measurements acquired thus
far. While there is an extensive amount of prior work investigating effective
practical approximations using variants of Shannon's entropy, the efficacy of
such policies heavily depends on the geometric distribution of objects in the
world. On the other hand, the principled approach of employing online POMDP
solvers is rendered impractical by the need to explicitly sample online from a
posterior distribution of world maps.
We present a novel data-driven imitation learning framework to efficiently
train information gathering policies. The policy imitates a clairvoyant oracle
- an oracle that at train time has full knowledge about the world map and can
compute maximally informative sensing locations. We analyze the learnt policy
by showing that offline imitation of a clairvoyant oracle is implicitly
equivalent to online oracle execution in conjunction with posterior sampling.
This observation allows us to obtain powerful near-optimality guarantees for
information gathering problems possessing an adaptive sub-modularity property.
As demonstrated on a spectrum of 2D and 3D exploration problems, the trained
policies enjoy the best of both worlds - they adapt to different world map
distributions while being computationally inexpensive to evaluate.Comment: Robotics Science and Systems, 201